Is the deconvolution layer the same as a convolutional layer?
نویسندگان
چکیده
1 In our CVPR 2016 paper [1], we proposed a novel network architecture to perform single image superresolution (SR). Most existing convolutional neural network (CNN) based superresolution methods [10,11] first upsample the image using a bicubic interpolation, then apply a convolutional network. We will refer to these types of networks as highresolution (HR) networks because the images are upsampled first. Instead, we feed the lowresolution (LR) input directly to a subpixel CNN as shown in Fig.1 : Figure 1: An illustration of the ESCPN framework where r denotes the upscaling ratio. Let denote the upscaling ratio e.g if the input LR image is then the output HR image will be r 1 × 1. We then output number of channels instead of one highresolution (HR) image and use periodic r × r r 2 shuffling to recreate the HR image. The exact details about how our efficient subpixel convolutional layer works can be found in the paper. We will refer to our network as a LR network. In this note, we want to focus on two aspects related to two questions most people asked us at CVPR when they saw this network. Firstly, how can channels magically become a HR image? And secondly, r 2 why are convolution in LR space a better choice? These are actually the key questions we tried to answer in the paper, but we were not able to go into as much depth and clarity as we would've liked given the page limit. To better answer these questions, we first discuss the relationships between the deconvolution layer in the form of the transposed convolution layer, the subpixel convolutional layer and our efficient subpixel convolutional layer, which we'll go through in Sec. 1 and Sec. 2. We will refer to our efficient subpixel convolutional layer as a convolutional layer in LR space to distinguish it from the common subpixel convolutional layer [5]. We will then show that for a fixed computational budget and complexity, a network with convolutions exclusively in LR space has more representation power at the same speed than a network that first upsamples the input in HR space.
منابع مشابه
A Convolutional Neural Network based on Adaptive Pooling for Classification of Noisy Images
Convolutional neural network is one of the effective methods for classifying images that performs learning using convolutional, pooling and fully-connected layers. All kinds of noise disrupt the operation of this network. Noise images reduce classification accuracy and increase convolutional neural network training time. Noise is an unwanted signal that destroys the original signal. Noise chang...
متن کاملCheckerboard artifact free sub-pixel convolution: A note on sub-pixel convolution, resize convolution and convolution resize
Convolutional neural networks (CNNs) are a popular and highly performant choice for pixel-wise dense prediction or generation. One of the commonly required components in such CNNs is a way to increase the resolution of the network’s input. The lower resolution inputs can be, for example, low-dimensional noise vectors in image generation [7] or low resolution (LR) feature maps for network vis...
متن کاملAn Information-Theoretic Discussion of Convolutional Bottleneck Features for Robust Speech Recognition
Convolutional Neural Networks (CNNs) have been shown their performance in speech recognition systems for extracting features, and also acoustic modeling. In addition, CNNs have been used for robust speech recognition and competitive results have been reported. Convolutive Bottleneck Network (CBN) is a kind of CNNs which has a bottleneck layer among its fully connected layers. The bottleneck fea...
متن کاملGenerative Deep Deconvolutional Learning
A generative model is developed for deep (multi-layered) convolutional dictionary learning. A novel probabilistic pooling operation is integrated into the deep model, yielding efficient bottom-up (pretraining) and top-down (refinement) probabilistic learning. After learning the deep convolutional dictionary, testing is implemented via deconvolutional inference. To speed up this inference, a new...
متن کاملCompressive Sensing via Convolutional Factor Analysis
We solve the compressive sensing problem via convolutional factor analysis, where the convolutional dictionaries are learned in situ from the compressed measurements. An alternating direction method of multipliers (ADMM) paradigm for compressive sensing inversion based on convolutional factor analysis is developed. The proposed algorithm provides reconstructed images as well as features, which ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1609.07009 شماره
صفحات -
تاریخ انتشار 2016